# PubMed Pre-training
Biomistral 7B DARE AWQ QGS128 W4 GEMM
An open-source large language model for the medical domain based on the Mistral architecture, further pre-trained on PubMed Central texts
Large Language Model
Transformers

B
BioMistral
135
3
Biomistral 7B
Apache-2.0
BioMistral is an open-source large language model optimized for the medical domain based on the Mistral architecture, further pre-trained on PubMed Central open-access text data, supporting multilingual medical question-answering tasks.
Large Language Model
Transformers Supports Multiple Languages

B
BioMistral
22.59k
428
Medcpt Article Encoder
Other
MedCPT is a model capable of generating biomedical text embeddings, particularly suitable for semantic search (dense retrieval) tasks.
Text Embedding
Transformers

M
ncbi
14.37k
24
Bioelectra PICO
BioELECTRA is a biomedical domain-specific language model pre-trained based on the ELECTRA framework, setting performance records on various biomedical NLP tasks
Large Language Model
Transformers

B
kamalkraj
10.88k
8
Ner Disease Ncbi Bionlp Bc5cdr Pubmed
Apache-2.0
Named entity recognition model trained on NCBI Disease dataset and BC5CDR dataset, specialized in identifying disease entities in biomedical literature
Sequence Labeling
Transformers Supports Multiple Languages

N
raynardj
10.84k
11
Bioformer 8L
Apache-2.0
A lightweight BERT model specifically designed for biomedical text mining, running 3 times faster than BERT-base while achieving comparable or superior performance to BioBERT/PubMedBERT
Large Language Model
Transformers English

B
bioformers
164
7
Featured Recommended AI Models